Goto

Collaborating Authors

 british columbia


Clio-X: AWeb3 Solution for Privacy-Preserving AI Access to Digital Archives

Lemieux, Victoria L., Gil, Rosa, Molosiwa, Faith, Zhou, Qihong, Li, Binming, Garcia, Roberto, Cubillo, Luis De La Torre, Wang, Zehua

arXiv.org Artificial Intelligence

As archives turn to artificial intelligence to manage growing volumes of digital records, privacy risks inherent in current AI data practices raise critical concerns about data sovereignty and ethical accountability. This paper explores how privacy-enhancing technologies (PETs) and Web3 architectures can support archives to preserve control over sensitive content while still being able to make it available for access by researchers. We present Clio-X, a decentralized, privacy-first Web3 digital solution designed to embed PETs into archival workflows and support AI-enabled reference and access. Drawing on a user evaluation of a medium-fidelity prototype, the study reveals both interest in the potential of the solution and significant barriers to adoption related to trust, system opacity, economic concerns, and governance. Using Rogers' Diffusion of Innovation theory, we analyze the sociotechnical dimensions of these barriers and propose a path forward centered on participatory design and decentralized governance through a Clio-X Decentralized Autonomous Organization. By integrating technical safeguards with community-based oversight, Clio-X offers a novel model to ethically deploy AI in cultural heritage contexts.


Graphics4Science: Computer Graphics for Scientific Impacts

Chen, Peter Yichen, Guo, Minghao, Pfister, Hanspeter, Lin, Ming, Freeman, William, Huang, Qixing, Shen, Han-Wei, Matusik, Wojciech

arXiv.org Artificial Intelligence

Computer graphics, often associated with films, games, and visual effects, has long been a powerful tool for addressing scientific challenges--from its origins in 3D visualization for medical imaging to its role in modern computational modeling and simulation. This course explores the deep and evolving relationship between computer graphics and science, highlighting past achievements, ongoing contributions, and open questions that remain. We show how core methods, such as geometric reasoning and physical modeling, provide inductive biases that help address challenges in both fields, especially in data-scarce settings. To that end, we aim to reframe graphics as a modeling language for science by bridging vocabulary gaps between the two communities. Designed for both newcomers and experts, Graphics4Science invites the graphics community to engage with science, tackle high-impact problems where graphics expertise can make a difference, and contribute to the future of scientific discovery. Additional details are available on the course website: https://graphics4science.github.io


Why AI can't take over creative writing

AIHub

In 1948, the founder of information theory, Claude Shannon, proposed modelling language in terms of the probability of the next word in a sentence given the previous words. These types of probabilistic language models were largely derided, most famously by linguist Noam Chomsky: "The notion of'probability of a sentence' is an entirely useless one." In 2022, 74 years after Shannon's proposal, ChatGPT appeared, which caught the attention of the public, with some even suggesting it was a gateway to super-human intelligence. Going from Shannon's proposal to ChatGPT took so long because the amount of data and computing time used was unimaginable even a few years before. ChatGPT is a large language model (LLM) learned from a huge corpus of text from the internet.


Revealed: The formula for the perfect day - including a short shift at WORK

Daily Mail - Science & tech

In the search for happiness, having a good day every day is surely crucial. But when there are so many pursuits competing for our attention, sometimes it's difficult to know how much time to allocate for each one. Now, scientists in Canada claim to cracked the code for the perfect day – and surprisingly, it includes a short shift at work. According to the experts, the formula for the perfect day is six hours of family time, two hours spent with friends, 1.5 hour socialising, two hours exercising and one hour eating and drinking. Additionally, the perfect day should involve no more than six hours of work and less than 15 minutes commuting.


Impact of Data Patterns on Biotype identification Using Machine Learning

Yu, Yuetong, Ge, Ruiyang, Hacihaliloglu, Ilker, Rauscher, Alexander, Tam, Roger, Frangou, Sophia

arXiv.org Artificial Intelligence

Background: Patient stratification in brain disorders remains a significant challenge, despite advances in machine learning and multimodal neuroimaging. Automated machine learning algorithms have been widely applied for identifying patient subtypes (biotypes), but results have been inconsistent across studies. These inconsistencies are often attributed to algorithmic limitations, yet an overlooked factor may be the statistical properties of the input data. This study investigates the contribution of data patterns on algorithm performance by leveraging synthetic brain morphometry data as an exemplar. Methods: Four widely used algorithms-SuStaIn, HYDRA, SmileGAN, and SurrealGAN were evaluated using multiple synthetic pseudo-patient datasets designed to include varying numbers and sizes of clusters and degrees of complexity of morphometric changes. Ground truth, representing predefined clusters, allowed for the evaluation of performance accuracy across algorithms and datasets. Results: SuStaIn failed to process datasets with more than 17 variables, highlighting computational inefficiencies. HYDRA was able to perform individual-level classification in multiple datasets with no clear pattern explaining failures. SmileGAN and SurrealGAN outperformed other algorithms in identifying variable-based disease patterns, but these patterns were not able to provide individual-level classification. Conclusions: Dataset characteristics significantly influence algorithm performance, often more than algorithmic design. The findings emphasize the need for rigorous validation using synthetic data before real-world application and highlight the limitations of current clustering approaches in capturing the heterogeneity of brain disorders. These insights extend beyond neuroimaging and have implications for machine learning applications in biomedical research.


Benchmarking Histopathology Foundation Models for Ovarian Cancer Bevacizumab Treatment Response Prediction from Whole Slide Images

Mallya, Mayur, Mirabadi, Ali Khajegili, Farahani, Hossein, Bashashati, Ali

arXiv.org Artificial Intelligence

Bevacizumab is a widely studied targeted therapeutic drug used in conjunction with standard chemotherapy for the treatment of recurrent ovarian cancer. While its administration has shown to increase the progression-free survival (PFS) in patients with advanced stage ovarian cancer, the lack of identifiable biomarkers for predicting patient response has been a major roadblock in its effective adoption towards personalized medicine. In this work, we leverage the latest histopathology foundation models trained on large-scale whole slide image (WSI) datasets to extract ovarian tumor tissue features for predicting bevacizumab response from WSIs. Our extensive experiments across a combination of different histopathology foundation models and multiple instance learning (MIL) strategies demonstrate capability of these large models in predicting bevacizumab response in ovarian cancer patients with the best models achieving an AUC score of 0.86 and an accuracy score of 72.5%. Furthermore, our survival models are able to stratify high- and low-risk cases with statistical significance (p < 0.05) even among the patients with the aggressive subtype of high-grade serous ovarian carcinoma. This work highlights the utility of histopathology foundation models for the task of ovarian bevacizumab response prediction from WSIs. The high-attention regions of the WSIs highlighted by these models not only aid the model explainability but also serve as promising imaging biomarkers for treatment prognosis.


Can machine learning predict citizen-reported angler behavior?

Schmid, Julia S., Simmons, Sean, Lewis, Mark A., Poesch, Mark S., Ramazi, Pouria

arXiv.org Artificial Intelligence

Prediction of angler behaviors, such as catch rates and angler pressure, is essential to maintaining fish populations and ensuring angler satisfaction. Angler behavior can partly be tracked by online platforms and mobile phone applications that provide fishing activities reported by recreational anglers. Moreover, angler behavior is known to be driven by local site attributes. Here, the prediction of citizen-reported angler behavior was investigated by machine-learning methods using auxiliary data on the environment, socioeconomics, fisheries management objectives, and events at a freshwater body. The goal was to determine whether auxiliary data alone could predict the reported behavior. Different spatial and temporal extents and temporal resolutions were considered. Accuracy scores averaged 88% for monthly predictions at single water bodies and 86% for spatial predictions on a day in a specific region across Canada. At other resolutions and scales, the models only achieved low prediction accuracy of around 60%. The study represents a first attempt at predicting angler behavior in time and space at a large scale and establishes a foundation for potential future expansions in various directions.


Pricey college textbooks next on AI's hit list? Professor says ChatGTP could replace them

FOX News

A professor says AI chatbot software, such as ChatGPT, could restructure postsecondary education by replacing some textbooks and promoting critical thinking. AI software like ChatGPT could be used to replace some university textbooks, transforming higher education and demanding a greater focus on critical thinking, a college professor told Fox News. "Certainly for an introductory undergraduate course, ChatGPT could be used to produce excellent course materials," said Terence Day, a geography professor at Okanagan College in British Columbia. "Essentially, it could substitute for a textbook." AI chatbots could make curricula less rigid and drive down college costs, since students wouldn't have to shell out hefty sums for textbooks, according to Day.


The Woman Daring Us to Build a World Without Oil and Coal

Mother Jones

The United States is on the brink of its most consequential transformation since the New Deal. Read more about what it takes to decarbonize the economy, and what stands in the way, here. This story was originally published by Hakai Magazine and is reproduced here as part of our Climate Desk collaboration. Imagination is a powerful thing. Mary Shelley predicted organ transplantation in her novel Frankenstein, published in 1818.


New AI Model Predicts Cancer Patient Survival More Accurately Than Previous Methods

#artificialintelligence

Predicting cancer patient survival rates is a crucial aspect of cancer treatment and management. Accurately forecasting a patient's prognosis helps medical professionals make informed decisions about the most appropriate course of action and can also aid in the development of personalized treatment plans. Researchers from the University of British Columbia and BC Cancer have created an AI model that predicts cancer patient survival with greater accuracy and using more readily accessible data compared to previous methods. The AI model utilizes natural language processing (NLP), a field of AI that comprehends human language, to examine oncologists' notes taken following a patient's initial consultation. This is the first step in a cancer patient's journey after diagnosis.